No DBA? No Regret! Multi-Armed Bandits for Index Tuning of Analytical and HTAP Workloads With Provable Guarantees
نویسندگان
چکیده
Automating physical database design has remained a long-term interest in research due to substantial performance gains afforded by optimised structures. Despite significant progress, majority of today's commercial solutions are highly manual, requiring offline invocation administrators (DBAs). This status quo is untenable: identifying representative static workloads no longer realistic; and tools remain susceptible the query optimiser's cost misestimates. Furthermore, modern application environments like hybrid transactional analytical processing (HTAP) systems render modelling next impossible. We propose self-driving approach online index selection that does not depend on DBA optimiser, instead learns benefits viable structures through strategic exploration direct observation. view problem as one sequential decision making under uncertainty, specifically within bandit learning setting. Multi-armed bandits balance exploitation provably guarantee average converges policies optimal with perfect hindsight. Our comprehensive empirical evaluation against state-of-the-art tuning tool demonstrates up 75% speed-up 59% HTAP environments. Lastly, our framework outperforms Monte Carlo tree search (MCTS)-based providing 24% speed-up.
منابع مشابه
No Regret Bound for Extreme Bandits
Algorithms for hyperparameter optimization abound, all of which work well under different and often unverifiable assumptions. Motivated by the general challenge of sequentially choosing which algorithm to use, we study the more specific task of choosing among distributions to use for random hyperparameter optimization. This work is naturally framed in the extreme bandit setting, which deals wit...
متن کاملBounded regret in stochastic multi-armed bandits
We study the stochastic multi-armed bandit problem when one knows the value μ(⋆) of an optimal arm, as a well as a positive lower bound on the smallest positive gap ∆. We propose a new randomized policy that attains a regret uniformly bounded over time in this setting. We also prove several lower bounds, which show in particular that bounded regret is not possible if one only knows ∆, and bound...
متن کاملRegret Analysis of the Finite-Horizon Gittins Index Strategy for Multi-Armed Bandits
I prove near-optimal frequentist regret guarantees for the finite-horizon Gittins index strategy for multi-armed bandits with Gaussian noise and prior. Along the way I derive finite-time bounds on the Gittins index that are asymptotically exact and may be of independent interest. I also discuss computational issues and present experimental results suggesting that a particular version of the Git...
متن کاملSimple regret for infinitely many armed bandits
We consider a stochastic bandit problem with infinitely many arms. In this setting, the learner has no chance of trying all the arms even once and has to dedicate its limited number of samples only to a certain number of arms. All previous algorithms for this setting were designed for minimizing the cumulative regret of the learner. In this paper, we propose an algorithm aiming at minimizing th...
متن کاملBounded Regret for Finite-Armed Structured Bandits
We study a new type of K-armed bandit problem where the expected return of one arm may depend on the returns of other arms. We present a new algorithm for this general class of problems and show that under certain circumstances it is possible to achieve finite expected cumulative regret. We also give problemdependent lower bounds on the cumulative regret showing that at least in special cases t...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Knowledge and Data Engineering
سال: 2023
ISSN: ['1558-2191', '1041-4347', '2326-3865']
DOI: https://doi.org/10.1109/tkde.2023.3271664